Appears in Ecml-98 as a Research Note Pruning Decision Trees with Misclassiication Costs 1 Pruning Decision Trees

نویسندگان

  • P. Bradford
  • Clayton Kunz
  • Ron Kohavi
  • Cli Brunk
  • Carla E. Brodley
چکیده

We describe an experimental study of pruning methods for decision tree classi ers when the goal is minimizing loss rather than error. In addition to two common methods for error minimization, CART's cost-complexity pruning and C4.5's error-based pruning, we study the extension of cost-complexity pruning to loss and one pruning variant based on the Laplace correction. We perform an empirical comparison of these methods and evaluate them with respect to loss. We found that applying the Laplace correction to estimate the probability distributions at the leaves was bene cial to all pruning methods. Unlike in error minimization, and somewhat surprisingly, performing no pruning led to results that were on par with other methods in terms of the evaluation criteria. The main advantage of pruning was in the reduction of the decision tree size, sometimes by a factor of ten. While no method dominated others on all datasets, even for the same domain di erent pruning mechanisms are better for di erent loss matrices. 1 Pruning Decision Trees Decision trees are a widely used symbolic modeling technique for classi cation tasks in machine learning. The most common approach to constructing decision tree classi ers is to grow a full tree and prune it back. Pruning is desirable because the tree that is grown may over t the data by inferring more structure than is justi ed by the training set. Speci cally, if there are no con icting instances, the training set error of a fully built tree is zero, while the true error is likely to be larger. To combat this over tting problem, the tree is pruned back with the goal of identifying the tree with the lowest error rate on previously unobserved instances, breaking ties in favor of smaller trees (Breiman, Friedman, Olshen & Stone 1984, Quinlan 1993). Several pruning methods have been introduced in the literature, including cost-complexity pruning, reduced error pruning, pessimistic pruning, error-based pruning, penalty pruning, and MDL pruning. Historically, most pruning algorithms have been developed to minimize the expected error rate of the decision tree, assuming that classi cation errors have the same unit cost. Our objective in this paper is di erent than the above-mentioned studies. Instead of pruning to minimize error, we aim to study pruning algorithms with the goal of minimizing loss. In many practical applications one has a loss matrix associated with classi cation errors (Turney 1997), and pruning should be performed with respect to the loss matrix. Pruning for loss minimization can lead to di erent pruning behavior than does pruning for error minimization. In this paper, we investigate the behavior of several pruning algorithms. In addition to the two most common methods for error minimization, cost-complexity pruning (Breiman et al. 1984) and error-based pruning (Quinlan 1993), we study the extension of cost-complexity pruning to loss and a pruning variant based on the Laplace correction (Good 1965, Cestnik 1990). We perform an empirical comparison of these methods and evaluate them with respect to loss under two di erent matrices. We found that even for the same domain, di erent pruning mechanisms are better for di erent loss matrices. In addition, we found that adjusting the probability distributions at the leaves using the Laplace correction was bene cial to all methods. 2 The Pruning Algorithms and Evaluation Criteria Most pruning algorithms perform a post-order traversal of the tree, replacing a subtree by a single leaf node when the estimated error of the leaf replacing the subtree is lower than that of the subtree. The crux of the problem is to nd an honest estimate of error (Breiman et al. 1984), which is de ned as one that is not overly optimistic for a tree that was built to minimize errors in the rst place. The resubstitution error (error rate on the training set) does not provide a suitable estimate because a leaf-node replacing a subtree will never have fewer errors on the training set than the subtree. Two commonly used pruning algorithms for error minimization are C4.5's error-based pruning (Quinlan 1993) and CART's cost-complexity pruning (Breiman et al. 1984). We attempted to extend several error-based pruning to loss-based pruning. In some cases the extensions are obvious, but C4.5's error-based pruning based on con dence intervals does not extend easily. The naive idea of computing a con dence interval for each probability and computing the losses based on the upper bound of the interval for each class yields a distribution that does not add to one. Experimental results we made on some variants (e.g. normalizing the probabilities) did not perform well. Instead, we decided to use a Laplace-based pruning method. The Laplace-based pruning method we introduce here has a similar motivation to C4.5's error-based pruning. The Laplace correction method biases the probability towards a uniform distribution. Speci cally, if a node has m instances, c of which are from a given class, in a k-class problem, the probability assigned to the class is (c+1)=(m+ k) (Good 1965, Cestnik 1990). The Laplace correction makes the distribution at the leaves more uniform and less extreme. Given a node, we can compute the expected loss using the loss matrix. The expected loss of a subtree is the sum of expected loss of the leaves. The cost-complexity-pruning (CCP) algorithm used in CART penalizes the estimated error based on the subtree size. Speci cally, the error estimate assigned to a subtree is the resubstitution error plus a factor times the subtree size. An e cient search algorithm can be used to compute all the distinct values that change the tree size and the parameter is chosen to minimize the error on a holdout sample or using cross-validation. Once the optimal value of is found, the entire training set is used to grow the tree and it is pruned using this optimal value. In our experiments, we have used the holdout method, holding back 20% of the training set to estimate the best parameter. Cost complexity pruning extends naturally to loss matrices. Instead of estimating the error of a subtree, we estimate its loss (or cost), using the resubstitution loss and penalizing by the size of the tree times the factor as in error-based CCP. 3 A Comparison of Pruning Algorithms Our goal in designing these experiments was to understand which pruning methods work well when the decision tree classi er is evaluated on loss given a loss matrix. The basic decision tree growing algorithm is implemented in MLC ++ (Kohavi, Sommer eld & Dougherty 1996) and called MC4 (MLC ++ C4.5). It is a Top-Down Decision Tree induction algorithm very similar to C4.5. The algorithm grows the decision tree following the standard methodology of choosing the best attribute according to the gain-ratio evaluation criterion and stopping when a node has two or fewer instances. The trees are pruned using the following pruning algorithms: eb-fr Error-based pruning (C4.5) with probabilities estimated using frequency counts. eb-lc Error-based pruning with probabilities estimated using the Laplace correction. np-lc No-pruning with probabilities estimated using the Laplace correction. lp Laplace-based pruning with probabilities estimated using the Laplace correction. ccp-lc Cost-complexity pruning based on loss with probabilities estimated using the Laplace correction. The leaves of the trees are labeled with the class that minimizes expected loss based on the probability estimates at each leaf. In our initial experiments, the Laplace correction outperformed frequency counts in all variants. Therefore, excluding the basic method of error-based-pruning, all other pruning methods were run with the Laplace correction. Ten datasets were chosen from the UCI repository (Merz & Murphy 1997): adult (salary classi cation based on census bureau data), breast cancer diagnosis, chess, crx (credit), german (credit), pima diabetes, road (dirt), satellite images, shuttle, and vehicle. In choosing the datasets, we decided on the following desiderata: 1. Datasets should be two-class to make the evaluation easier. This desideratum was hard to satisfy and we resorted to converting several multi-class problems into two-class problems by choosing the least prevalent class as the goal class. 2. Datasets should not have too many unknowns. To avoid another factor in this evaluation, we removed all instances with unknown values from the les. 3. The standard error of the estimated loss should be small. This was very important because with loss matrices the standard deviations of the estimates can be large. We therefore decided to require at least 500 instances and train on only 25% of the data, leaving the remaining instances for testing. We wanted to test the following hypotheses: 1. The Laplace correction for estimating probabilities at the leaves leads to lower loss than frequency counts. 2. Considering the loss matrix during pruning leads to lower loss than pruning based on errors. For all datasets we trained on 25% of the data and tested on 75% of the data, repeating the process 10 times. We compared performance of the pruning algorithms on two di erent loss matrices, which respectively set a loss of 10 and 100 for misclassifying the less frequent of the two classes. This was done to simulate real-world scenarios in which the less frequent class is the important class. Experiments were also done with the losses reversed, with similar conclusions to those shown below. The results are displayed as graphs showing the average loss for the ten les as bars using the scale on the left, and the average relative loss as X-symbols with the scale on the right. The relative losses are computed as the ratio between the loss of the pruning method and eb-fr, our baseline method. These ratios are then averaged across the ten datasets to create summary graphs. In cases for which the losses are small, the ratio is a better indicator of performance. The average losses and average relative losses for the two loss matrices are shown in Figure 1. The following observations can be made: 1. Error-based pruning with frequency counts performs the worst. 2. The Laplace-based pruning (lp) performs the best on the 10 to 1 loss matrix and is comparable to the best on the 100 to 1 loss matrix. 3. No-pruning (np-lc) performs surprisingly well on both loss matrices! 4. Cost-complexity pruning (ccp-lc) is slightly inferior to no-pruning, but better than error-based pruning (eb) on the 100 to 1 loss matrix. 5. Tree sizes were radically di erent. The average tree sizes for the 10 to 1 loss matrix are: ccp(47), eb(118), lp(382), and np(670). Cost-complexity pruning was by far the smallest, con rming the observation by Oates & Jensen (1997) for error minimization. Average loss with 10:1 loss ratio Absolute Relative np-lc lp ccp-lc eb-lc eb-fr 0.0 0.1 0.2 0.3 0.4 0.5 0.0 0.2 0.4 0.6 0.8 1.0 Average loss with 100:1 loss ratio Absolute Relative np-lc lp ccp-lc eb-lc eb-fr 0 1 2 3 0.0 0.2 0.4 0.6 0.8 1.0 Fig. 1. Average absolute and relative losses for the di erent algorithms and for a 10 to 1 loss matrix (left) and a 100 to 1 matrix (right). Our hypothesis that the Laplace correction for estimating probabilities at the leaves outperforms frequency counts was con rmed. It was also con rmed for the np and ccp pruning methods when they were run with frequency counts (results not shown). Interestingly, no-pruning performed very well, suggesting that when we have loss matrices and when tree size is not important, pruning need not be done if the Laplace correction is used. This result di ers from error minimization, where pruning was consistently shown to help. Pruning based on loss matrices performed better than pruning based on error for frequency counts for all methods. This result (for frequency counts) has been observed previously for reduced error/cost pruning (Draper, Brodley & Utgo 1994). When the Laplace correction was used, pruning with loss matrices performed better than error-based pruning (eb-lc) for the 100:1 (ccp-lc, lp) but there was no signi cant di erence for the 10:1 loss matrix. For each pruning method, applying the Laplace correction improved performance on average. Only in a few cases did the Laplace correction lead to a higher distribution MSE (mean-squared-error) than frequency counts. The distribution MSE was similar for all the Laplace correction algorithms. The main di erence between the pruning algorithms was in the tree size. The average tree sizes were ccp(27), eb(118), lp(280), and np(670).)

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Appears in Ecml-98 as a Research Note a Longer Version Is Available as Ece Tr 98-3, Purdue University Pruning Decision Trees with Misclassiication Costs 1 Pruning Decision Trees

We describe an experimental study of pruning methods for decision tree classiiers when the goal is minimizing loss rather than error. In addition to two common methods for error minimization, CART's cost-complexity pruning and C4.5's error-based pruning, we study the extension of cost-complexity pruning to loss and one pruning variant based on the Laplace correction. We perform an empirical com...

متن کامل

Simplifying Decision Trees by Pruning and Grafting: New Results (Extended Abstract)

This paper presents some empirical results on simplification methods of decision trees induced from data. We observe that those methods exploiting an independent pruning set do not perform uniformly better than the others. Furthermore, a clear defmition of bias towards overpruning and underpraning is exploited in order to interpret empirical data concerning the size of the simplified trees.

متن کامل

A New Pruning Method for Solving Decision Trees and Game Trees

The main goal of this paper is to describe a new pruning method for solving decision trees and game trees. The pruning method for decision trees suggests a slight variant of decision trees that we call scenario trees. In scenario trees, we do not need a conditional probability for each edge emanating from a chance node. Instead, we require a joint probability for each path from the root node to...

متن کامل

Cost-sensitive Decision Trees with Post-pruning and Competition for Numeric Data

Decision tree is an effective classification approach in data mining and machine learning. In some applications, test costs and misclassification costs should be considered while inducing decision trees. Recently, some cost-sensitive learning algorithms based on ID3, such as CS-ID3, IDX, ICET and λ-ID3, have been proposed to deal with the issue. In this paper, we develop a decision tree algorit...

متن کامل

The Effect of Fruit Trees Pruning Waste Biochar on some Soil Biological Properties under Rhizobox Conditions

The pyrolysis of fruit trees Pruning waste to be converted to biochar with microbial inoculation is a strategy improving the biological properties in calcareous soils. In order to investigate the biochar effect on some soil biological properties of the soil in the presence of microorganisms, a factorial experiment was carried out in a completely randomized design in the rhizobox under greenhous...

متن کامل

On the Boosting Pruning Problem

Boosting is a powerful method for improving the predic-tive accuracy of classiiers. The AdaBoost algorithm of Freund and Schapire has been successfully applied to many domains 2, 10, 12] and the combination of AdaBoost with the C4.5 decision tree algorithm has been called the best oo-the-shelf learning algorithm in practice. Unfortunately , in some applications, the number of decision trees req...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1998